Optimum Follow the Leader Algorithm
نویسندگان
چکیده
Consider the following setting for an on-line algorithm (introduced in [FS97]) that learns from a set of experts: In trial t the algorithm chooses an expert with probability pi. At the end of the trial a loss vector 1 L ∈ [0, R] for the n experts is received and an expected loss of ∑ i p t iL t i is incurred. A simple algorithm for this setting is the Hedge algorithm which uses the probabilities pi ∼ exp−ηL <t i . This algorithm and its analysis is a simple reformulation of the randomized version of the Weighted Majority algorithm (WMR) [LW94] which was designed for the absolute loss. The total expected loss of the algorithm is close to the total loss of the best expert L∗ = mini L ≤T i . That is, when the learning rate is optimally tuned based on L∗, R and n, then the total expected loss of the Hedge/WMR algorithm is at most
منابع مشابه
The Price of Optimum in Stackelberg Games
Consider a system M of parallel machines, each with a strictly increasing and differentiable load dependent latency function. The users of such a system are of infinite number and act selfishly, routing their infinitesimally small portion of the total flow r they control, to machines of currently minimum delay. It is well known that such selfishness if modeled by a noncooperative game may yield...
متن کاملOptimization in Uncertain and Complex Dynamic Environments with Evolutionary Methods
In the real world, many of the optimization issues are dynamic, uncertain, and complex in which the objective function or constraints can be changed over time. Consequently, the optimum of these issues is changed nonlinearly. Therefore, the optimization algorithms not only should search the global optimum value in the space but also should follow the path of optimal change in dynamic environmen...
متن کاملOnline Linear Optimization via Smoothing
We present a new optimization-theoretic approach to analyzing Follow-the-Leader style algorithms, particularly in the setting where perturbations are used as a tool for regularization. We show that adding a strongly convex penalty function to the decision rule and adding stochastic perturbations to data correspond to deterministic and stochastic smoothing operations, respectively. We establish ...
متن کاملFollow the Moving Leader in Deep Learning
Deep networks are highly nonlinear and difficult to optimize. During training, the parameter iterate may move from one local basin to another, or the data distribution may even change. Inspired by the close connection between stochastic optimization and online learning, we propose a variant of the follow the regularized leader (FTRL) algorithm called follow the moving leader (FTML). Unlike the ...
متن کاملMulti-Objective Particle Swarm Optimization Algorithms – A Leader Selection Overview
Multi Objective Optimization (MOO) problem involves simultaneous minimization or maximization of many objective functions. Various MOO algorithms have been introduced to solve the MOO problem. Traditional gradient-based techniques are one of the methods used to solve MOO problems. However, in the traditional gradient-based technique only one solution is generated. Thus, an alternative approach ...
متن کاملObstacle-free Control of the Hyper-redundant Nasa Inspection Manipulator
This paper presents a follow-the-leader algorithm for serpentine control of hyper-redundant manipulators. Given an obstacle-free trajectory for the manipulator tip generated by a path planner or teleoperation, the follow-the-leader algorithm ensures whole-arm collision avoidance by forcing ensuing links to follow the same trajectory. The algorithm requires two steps to place the tip on each tra...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2005